Text Embedding with Advanced Recurrent Neural Model

نویسندگان

  • Shuochao Yao
  • Shiqi Zhou
چکیده

Embedding method has become a popular way to handle unstructured data, such as word and text. Word embedding, providing computational-friendly representations for word similarity, is almost be one of the standard solutions for various text mining tasks. Lots of recent studies focusing on word embedding try to generate a more comprehensive representation for each word that incorporating task-specific information. Yet compared with massive studies on word embedding, text embedding is still facing a lot of challenging problems. As “bag of embeddings” method is mainly used to generate text embedding, how to handle the order information in sentence becomes an important problem to be solved. Distributed representation, i.e. embedding, provides the possibility to encode the order information in a relative low-dimensional space, freeing the constraint of analyzing exponential states in sentence with traditional onehot representation. One promising approach to handle order information is recurrent neural network. Yet the original recurrent neural network and its extension, Long-Short Term Memory (LSTM) and Gated Recurrent Unit (GRU), still suffer the long-term gradient vanishing problem. Therefore long-term dependency or relationship is still hard to handle. In this project, we prosed two models, attention model and explicit memory model, that explicitly consider longterm dependency that try to solve the long-term dependency problem in text embedding. Experiments on real datasets illustrate their potential capacities to handle long-term dependency in text embedding task and related applications.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Joint Semantic Vector Representation Model for Text Clustering and Classification

Text clustering and classification are two main tasks of text mining. Feature selection plays the key role in the quality of the clustering and classification results. Although word-based features such as term frequency-inverse document frequency (TF-IDF) vectors have been widely used in different applications, their shortcoming in capturing semantic concepts of text motivated researches to use...

متن کامل

A New Document Embedding Method for News Classification

Abstract- Text classification is one of the main tasks of natural language processing (NLP). In this task, documents are classified into pre-defined categories. There is lots of news spreading on the web. A text classifier can categorize news automatically and this facilitates and accelerates access to the news. The first step in text classification is to represent documents in a suitable way t...

متن کامل

Learning text representation using recurrent convolutional neural network with highway layers

Recently, the rapid development of word embedding and neural networks has brought new inspiration to various NLP and IR tasks. In this paper, we describe a staged hybrid model combining Recurrent Convolutional Neural Networks (RCNN) with highway layers. The highway network module is incorporated in the middle takes the output of the bidirectional Recurrent Neural Network (Bi-RNN) module in the ...

متن کامل

Articulatory movement prediction using deep bidirectional long short-term memory based recurrent neural networks and word/phone embeddings

Automatic prediction of articulatory movements from speech or text can be beneficial for many applications such as speech recognition and synthesis. A recent approach has reported stateof-the-art performance in speech-to-articulatory prediction using feed forward neural networks. In this paper, we investigate the feasibility of using bidirectional long short-term memory based recurrent neural n...

متن کامل

Dual-Path Convolutional Image-Text Embedding

This paper considers the task of matching images and sentences. The challenge consists in discriminatively embedding the two modalities onto a shared visual-textual space. Existing work in this field largely uses Recurrent Neural Networks (RNN) for text feature learning and employs off-the-shelf Convolutional Neural Networks (CNN) for image feature extraction. Our system, in comparison, differs...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2016